Estimation bounds and sharp oracle inequalities of regularized procedures with Lipschitz loss functions
نویسندگان
چکیده
We obtain estimation error rates and sharp oracle inequalities for regularization procedures of the form f̂ ∈ argmin f∈F ( 1 N N ∑ i=1 `(f(Xi), Yi) + λ ‖f‖ ) when ‖·‖ is any norm, F is a convex class of functions and ` is a Lipschitz loss function satisfying a Bernstein condition over F . We explore both the bounded and subgaussian stochastic frameworks for the distribution of the f(Xi)’s, with no assumption on the distribution of the Yi’s. The general results rely on two main objects: a complexity function, and a sparsity equation, that depend on the specific setting in hand (loss ` and norm ‖·‖). As a proof of concept, we obtain minimax rates of convergence in the following problems: 1) matrix completion with any Lipschitz loss function, including the hinge and logistic loss for the so-called 1-bit matrix completion instance of the problem, and quantile losses for the general case, which enables to estimate any quantile on the entries of the matrix; 2) logistic LASSO and variants such as the logistic SLOPE; 3) kernel methods, where the loss is the hinge loss, and the regularization function is the RKHS norm.
منابع مشابه
Non-Asymptotic Oracle Inequalities for the High-Dimensional Cox Regression via Lasso.
We consider finite sample properties of the regularized high-dimensional Cox regression via lasso. Existing literature focuses on linear models or generalized linear models with Lipschitz loss functions, where the empirical risk functions are the summations of independent and identically distributed (iid) losses. The summands in the negative log partial likelihood function for censored survival...
متن کاملar X iv : 0 80 3 . 28 39 v 1 [ m at h . ST ] 1 9 M ar 2 00 8 AGGREGATION BY EXPONENTIAL WEIGHTING , SHARP ORACLE INEQUALITIES AND SPARSITY
We study the problem of aggregation under the squared loss in the model of regression with deterministic design. We obtain sharp PAC-Bayesian risk bounds for aggregates defined via exponential weights, under general assumptions on the distribution of errors and on the functions to aggregate. We then apply these results to derive sparsity oracle inequalities.
متن کاملAdaptive Dantzig density estimation
This paper deals with the problem of density estimation. We aim at building an estimate of an unknown density as a linear combination of functions of a dictionary. Inspired by Candès and Tao’s approach, we propose an l1-minimization under an adaptive Dantzig constraint coming from sharp concentration inequalities. This allows to consider a wide class of dictionaries. Under local or global coher...
متن کاملLinear and convex aggregation of density estimators
We study the problem of learning the best linear and convex combination of M estimators of a density with respect to the mean squared risk. We suggest aggregation procedures and we prove sharp oracle inequalities for their risks, i.e., oracle inequalities with leading constant 1. We also obtain lower bounds showing that these procedures attain optimal rates of aggregation. As an example, we con...
متن کاملTo “ General Non - Exact Oracle Inequalities for Classes with a Subexponential Envelope ”
We apply Theorem A to the problem of Convex aggregation and show that the optimal rate of Convex aggregation for non-exact oracle inequalities is much faster than the optimal rate for exact oracle inequalities. We apply Theorem B to show that regularized procedures based on a nuclear norm criterion satisfy oracle inequalities with a residual term that decreases like 1/n for every Lq-loss functi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017